Bayesian Exploration: Incentivizing Exploration in Bayesian Games
نویسندگان
چکیده
In a wide range of recommendation systems, self-interested individuals (“agents”) make decisions over time, using information revealed by other agents in the past, and producing that may help future. Each agent would like to exploit best action given current but prefer previous explore various alternatives collect information. A social planner, means well-designed policy, can incentivize balance exploration exploitation order maximize welfare or some objective. The policy be modeled as multiarmed bandit algorithm under Bayesian incentivecompatibility (BIC) constraints. This line work has received considerable attention “economics computation” community. Although prior work, planner interacts with single at present paper allows affect one another directly shared environment. now face two sources uncertainty: what is environment, do? We focus on “explorable” actions: those recommended BIC policy. show how principal identify all such actions.
منابع مشابه
RNA-Seq Bayesian Network Exploration of Immune System in Bovine
Background: The stress is one of main factors effects on production system. Several factors (both genetic and environmental elements) regulate immune response to stress. Objectives: In order to determine the major immune system regulatory genes underlying stress responses, a learning Bayesian network approach for those regulatory genes was applied to RNA-...
متن کاملModel based Bayesian Exploration
Reinforcement learning systems are often concerned with balancing exploration of untested actions against exploitation of actions that are known to be good. The benefitof exploration can be estimated using the classical notion of Value of Information — the expected improvement in future decision quality arising from the information acquired by exploration. Estimating this quantity requires an a...
متن کاملBayesian Reinforcement Learning with Exploration
We consider a general reinforcement learning problem and show that carefully combining the Bayesian optimal policy and an exploring policy leads to minimax sample-complexity bounds in a very general class of (history-based) environments. We also prove lower bounds and show that the new algorithm displays adaptive behaviour when the environment is easier than worst-case.
متن کاملBayesian Exploration for Mobile Robots
This work addresses the problem of robot exploration. That is, the task of automatically learning a map of the environment which is useful for mobile robot navigation and localization. The exploration mechanism is intended to be applicable to an arbitrary environment, and is independent of the particular representation of the world. We take an information-theoretic approach and avoid the use of...
متن کاملTutorial: Incentivizing and Coordinating Exploration
An increasing variety of platforms and markets rely on the activities of self-interested agents to explore a space of alternatives by engaging in the (often costly) process of acquiring information about those alternatives. The cost of exploration may be direct, such as paying to interview job candidates prior to making a hire or to visit colleges prior to deciding on a school, or it may be an ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Operations Research
سال: 2022
ISSN: ['1526-5463', '0030-364X']
DOI: https://doi.org/10.1287/opre.2021.2205